Towards Faster Training of Global Covariance Pooling Networks by Iterative Matrix Square Root Normalization

نویسندگان

  • Peihua Li
  • Jiangtao Xie
  • Qilong Wang
  • Zilin Gao
چکیده

Global covariance pooling in convolutional neural networks has achieved impressive improvement over the classical first-order pooling. Recent works have shown matrix square root normalization plays a central role in achieving state-of-the-art performance. However, existing methods depend heavily on eigendecomposition (EIG) or singular value decomposition (SVD), suffering from inefficient training due to limited support of EIG and SVD on GPU. Towards addressing this problem, we propose an iterative matrix square root normalization method for fast end-toend training of global covariance pooling networks. At the core of our method is a meta-layer designed with loopembedded directed graph structure. The meta-layer consists of three consecutive nonlinear structured layers, which perform pre-normalization, coupled matrix iteration and post-compensation, respectively. Our method is much faster than EIG or SVD based ones, since it involves only matrix multiplications, suitable for parallel implementation on GPU. Moreover, the proposed network with ResNet architecture can converge in much less epochs, further accelerating network training. On large-scale ImageNet, we achieve competitive performance superior to existing counterparts. By finetuning our models pre-trained on ImageNet, we establish state-of-the-art results on three challenging finegrained benchmarks. The source code and network models will be available at http://www.peihuali.org/iSQRT-COV.

برای دانلود رایگان متن کامل این مقاله و بیش از 32 میلیون مقاله دیگر ابتدا ثبت نام کنید

ثبت نام

اگر عضو سایت هستید لطفا وارد حساب کاربری خود شوید

منابع مشابه

Improved Bilinear Pooling with CNNs

Bilinear pooling of Convolutional Neural Network (CNN) features [22, 23], and their compact variants [10], have been shown to be effective at fine-grained recognition, scene categorization, texture recognition, and visual question-answering tasks among others. The resulting representation captures second-order statistics of convolutional features in a translationally invariant manner. In this p...

متن کامل

MoNet: Moments Embedding Network

Bilinear pooling has been recently proposed as a feature encoding layer, which can be used after the convolutional layers of a deep network, to improve performance in multiple vision tasks. Different from conventional global average pooling or fully connected layer, bilinear pooling gathers 2nd order information in a translation invariant fashion. However, a serious drawback of this family of p...

متن کامل

Computing the Matrix Geometric Mean of Two HPD Matrices: A Stable Iterative Method

A new iteration scheme for computing the sign of a matrix which has no pure imaginary eigenvalues is presented. Then, by applying a well-known identity in matrix functions theory, an algorithm for computing the geometric mean of two Hermitian positive definite matrices is constructed. Moreover, another efficient algorithm for this purpose is derived free from the computation of principal matrix...

متن کامل

Comparison results on the preconditioned mixed-type splitting iterative method for M-matrix linear systems

Consider the linear system Ax=b where the coefficient matrix A is an M-matrix. In the present work, it is proved that the rate of convergence of the Gauss-Seidel method is faster than the mixed-type splitting and AOR (SOR) iterative methods for solving M-matrix linear systems. Furthermore, we improve the rate of convergence of the mixed-type splitting iterative method by applying a preconditio...

متن کامل

L1-Norm Batch Normalization for Efficient Training of Deep Neural Networks

Batch Normalization (BN) has been proven to be quite effective at accelerating and improving the training of deep neural networks (DNNs). However, BN brings additional computation, consumes more memory and generally slows down the training process by a large margin, which aggravates the training effort. Furthermore, the nonlinear square and root operations in BN also impede the low bit-width qu...

متن کامل

ذخیره در منابع من


  با ذخیره ی این منبع در منابع من، دسترسی به آن را برای استفاده های بعدی آسان تر کنید

عنوان ژورنال:
  • CoRR

دوره abs/1712.01034  شماره 

صفحات  -

تاریخ انتشار 2017